Cerebral processing of emotional prosody: a network model based on fMRI studies

نویسندگان

  • Dirk Wildgruber
  • Thomas Ethofer
  • Benjamin Kreifelts
  • Didier Grandjean
چکیده

Comprehension of information conveyed by emotional tone of speech is highly important for successful social interactions. Regarding the underlying neurobiological mechanisms, successive steps of cerebral processing involving auditory analysis within the temporal cortex and evaluative judgements within the frontal lobes have been differentiated (Schirmer & Kotz, 2006; Wildgruber et al., 2006). To further disentangle the impact of stimulus properties and appraisal levels a series of fMRI studies has been performed. The results of these studies indicate a strong association of cerebral responses and acoustic properties of the stimuli in some regions (stimulusdriven effects), whereas other areas showed modulation of activation linked to the focussing of attention to specific task components (task-dependent effects). Based on these findings a refined model of prosody processing and cross-modal integration of emotional signals from face and voice is postulated. 1. Stimulus-driven effects Enhanced activation within the middle section of the superior temporal cortex (mid-STC) has been observed in response to human voices as compared to other acoustic signals (Belin, 2000). Regarding emotional prosody, Grandjean and collaborators (Grandjean, 2005) have demonstrated increasing responses within this region related to anger prosody independent of the participant’s spatial attention during a dichotic listening task (Fig. 1). Figure 1: Activations related to angry prosody independent of spatial attention (red) and responses linked to spatial attention (green) rendered on the cortex of a standard brain. Activation within the mid-STC is encircled with blue line (adapted from Grandjean et al., 2005). In another fMRI study emotional adjectives spoken in happy, angry or neutral intonation were presented to healthy subjects. At semantic level, stimuli comprised unpleasant, pleasant and neutral words. Emotions conveyed by word content and prosody were varied independently. In separate runs, participants either rated the emotional valence of verbal content or the emotional valence of prosody. Independent of the task, enhanced activation within the mid-STC was associated to increasing intensity of emotional prosody (Ethofer et al., 2006a, Fig. 2). Figure 2: Stimulus-driven effects. Regions characterized by a significant linear relationship between hemodynamic responses and prosodic emotional intensity independent of the task (evaluation of word content or prosody, adapted from Ethofer et al., 2006a). In a further study passive listening to adjectives and substantives with neutral word content, spoken in five different emotional intonations (happy, neutral, fearful, angry and alluring) was evaluated. All four emotional categories induced stronger responses within the right mid-STC than neutral stimuli (Fig. 3). These responses were significantly correlated with several acoustic parameters (stimulus duration, mean intensity, mean pitch and pitch variability). Figure 3: Activation during passive listening to emotional prosody (emotional prosody > neutral prosody). (a) Projection of significant activation rendered on the right hemisphere of a standard brain. (b) Transversal slice showing activation of mid-STC and thalamus. (c) Effect size in right mid-STC for the different prosodic categories. (d) Gender dependent responses during perception of alluring prosody (m = male listener, f = female listener). Error bars represent standard errors of the mean. L R A multiple regression analysis revealed that none of the parameters alone could explain the observed activation pattern. The conjoint effect of acoustic parameters, however, sufficiently explained increase of responses within the right mid-STC. Therefore, an important contribution of this area to the integration of several acoustic parameters into an emotional percept has been suggested (Wiethoff et al., 2008). Moreover, an analysis of interaction effects between gender of the speaker and the listener revealed a cross-gender interaction with increasing responses to the voice of the opposite sex in male and female subjects. This effect was confined to an alluring tone of speech in behavioural data as well as hemodynamic responses within the mid-STC (Fig. 3d). The response pattern of the mid-STC, thus, indicates a particular sensitivity to emotional voices with a high behavioural relevance for the listener (Ethofer et al., 2007). 2. Task-dependent effects If attention of the participants is explicitly directed towards the emotional auditory stimulus by specific task instructions further cerebral regions are activated. When subjects explicitly judged the emotional valence conveyed by prosody, for example, increased responses within the right posterior STC (post-STC) and the bilateral inferior frontal cortex were observed as compared to explicit evaluation of the verbal content of identical stimuli (Ethofer et al., 2006b, Fig. 4a). Figure 4: Task-dependent effects. (a) Increasing cerebral responses during explicit evaluation of emotional prosody at word level as compared to rating of emotional word content rendered on the cortex of a standard brain. Stimulus-driven activation during the same experiment is shown in Fig. 2. (b) Task-dependent effects at sentence level. Significant activations during categorization of emotional prosody as compared to the phonetic control task. Evaluation of emotional prosody yielded activation within the right post-STC and inferior frontal cortex. During another experiment semantically neutral sentences (i.e. „The visitor reserved a room for Thursday“), spoken by actors in different emotional tones (happy, fearful, angry, sad, disgusted) served as stimulus material (Wildgruber et al., 2005). Participants either had to name the emotional category expressed by prosody or they had to perform a phonetic control task (identification of the vowel following the first “a” in the sentence). In accordance with the findings during processing of single words, explicit evaluation of emotional prosody at sentence level was associated with activation of the right post-STC and the inferior frontal cortex (Fig. 4b). A different pattern of task-dependent effects was observed, when subjects were asked to voluntarily modulate spatial attention while performing a gender discrimination task during a dichotic listening experiment. Anger prosody presented at the to-be-attended side yielded increasing activation within the medial orbitofrontal cortex (OFC) and the medial occipital cortex as compared to presentation of identical stimuli at the to-be-ignored ear (Sander et al., 2005, Fig. 5). The right amygdala and the bilateral mid-STC, in contrast, responded to anger prosody irrespective of the direction of spatial attention. Figure 5: Spatial attention dependent effects. Meaningless utterances, pronounced with either angry or neutral prosody were presented in a dichotic listening paradigm. Participants were asked to selectively attend to either the left or right ear and performed a gender-decision on the voice heard on the target side. Increasing activation to the angry voice presented on the to-be-attended as compared to the to-be-ignored ear was observed within the medial orbitofrontal cortex (a) and the medial occipital cortex (b) (adapted form Sander et al., 2005). Besides emotional information, speech prosody can also convey information about linguistic meaning (e.g. determining if a sentence is a statement, a question or a command). Experimental findings indicate that the contribution of the right and the left cerebral hemisphere to extraction of acoustic signals depends upon specific stimulus properties (Wildgruber et al., 2006; Reiterer et al., 2005, 2008). According to the acoustic-lateralization hypothesis slow changes of acoustic signals (e.g. modulations of prosody) are processed within the right hemisphere, whereas the left hemisphere is better suited to process rapid changes of acoustic signals (e.g. differentiation of speech sounds at the b identification of emotional prosody (vs. vowel identification) level of syllables or phonems). To further disentangle the impact of functional aspects (emotional vs. linguistic) from basic auditory processing, evaluation of emotional and linguistic prosody was compared using acoustically highly controlled stimuli. To this end, the intonation-contour of a semantically neutral sentence (“the scarf is in the chest”) was systematically manipulated by digital resynthesis. Participants were asked to differentiate pairs of these sentences either with respect to their emotional arousal (“which of the two sentences sounds more exited?”) or their sentence focus (“which of the two sentences is better suited to answer the question: where is the scarf?”). As compared to rest both tasks yielded bilateral frontotemporal activation including rightward lateralization at the level of the post-STC (Wildgruber et al., 2004). Analysis of task-effects, however, revealed activation of bilateral orbitofrontal cortices linked to emotional evaluation (Fig. 6). Figure 6: Task-dependent activation during evaluation of emotional prosody as compared to evaluation of linguistic prosody. Significant responses were rendered on the cortex of a standard brain (left) and a transversal slice (right) at the level of the highest activated voxel. 3. Model of emotional prosody processing An analysis of effective connectivity revealed that the right post-STC is the most likely input region into the network of areas characterised by task-dependent activation (Ethofer et al., 2006b). This finding is in line with the assumption that this area subserves representation of meaningful prosodic sequences and receives direct input from primary and secondary acoustic regions. Moreover, the connectivity analysis indicated a flow of information along parallel projections from the right post-STC to the bilateral inferior frontal cortex. Taken together these findings indicate multiple successive processing stages, during recognition of emotional prosody following representation within the primary auditory cortex (A1). The first step, extraction of supra-segmental acoustic information, is associated with activation of predominantly right hemispheric primary and secondary acoustic regions. The second step, representation of meaningful supra-segmental acoustic sequences, is linked to posterior aspects of the right superior temporal sulcus. The third step, emotional judgment, is linked to the bilateral inferior-frontal cortex (IFC). Within this network, the projection from primary acoustic regions (A1) to the secondary acoustic representation within the mid-STC seems to be predominantly stimulus-driven (bottom-up effect), whereas further projections to the post-STC and the IFC depend upon focussing of attention towards explicit emotional evaluation (top-down effects). Considering implicit processing of emotional prosody, task-independent activation of the amygdale has been observed (Sander et al., 2005; Ethofer et al., 2008). Additionally, top-down modulation depending on spatial attention has been observed within the medial OFC. Therefore, this region might contribute to processing of emotionally relevant information within the scope of attention even if no explicit emotional judgement is required. Moreover, the observed activation pattern within the OFC seems to indicate that this area might contribute to the inhibition of potentially distracting information from unattended sources. Presumably, the medial OFC receives its input from subcortical limbic regions such as the amygdala (Fig. 7). Figure 7: Model of emotional prosody processing. Processing steps during explicit evaluation (following signal transmission via thalamus and A1): 1) Secondary auditory representation (mid-STC) 2) Identification of prosodic sequences (post-STC) 3) Explicit emotional evaluation (IFC / OFC) Flow of information during implicit processing 1) Thalamus 2) Amygdala 3) Medial orbitofrontal cortex (OFC) Bottom-up modulation (stimulus-driven) is indicated with red arrows. Top-down modulation (task-dependent) is indicated with green arrows. A1 = primary auditory cortex, mid-STC = middle section of the superior temporal cortex, post-STC = posterior section of the STC, IFC = inferior frontal cortex, OFC = orbitofrontal cortex. 4. Cross modal integration of emotional signals Considering bimodal integration, evaluation of audiovisual emotional signals yielded increased activation within the bilateral post-STC and the right thalamus as compared to either unimodal stimulation (Ethofer et al., 2006d, Kreifelts et al., 2007). Moreover, enhanced connectivity between the bilateral post-STC and voice-sensitive (mid-STC) as well as face-sensitive (fusiform face area) regions during processing of multimodal signals has been observed, which possibly depicts the mechanism of bimodal binding (Kreifelts et al, 2007). Presumably, the formation of such multimodal associations within the post-STC might also contribute to correct understanding of unimodal communicative signals. A1

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Cerebral processing of linguistic and emotional prosody: fMRI studies.

During acoustic communication in humans, information about a speaker's emotional state is predominantly conveyed by modulation of the tone of voice (emotional or affective prosody). Based on lesion data, a right hemisphere superiority for cerebral processing of emotional prosody has been assumed. However, the available clinical studies do not yet provide a coherent picture with respect to inter...

متن کامل

euronal synchronization during auditory emotional perception

The brain regions subserving the processing of emotional prosody have been investigated using functional magnetic resonance imaging (fMRI) and electroencephalography (EEG) but few studies investigated direct electrical neuronal activity using intracranial recordings in humans. Previous studies have shown the involvement of the amygdala and the orbito-frontal cortex (OFC) in the processing of em...

متن کامل

Speech Prosody 2008_Grandjean&Scherer

The brain regions subserving the processing of emotional prosody have been investigated using functional magnetic resonance imaging (fMRI) and electroencephalography (EEG) but few studies investigated direct electrical neuronal activity using intracranial recordings in humans. Previous studies have shown the involvement of the amygdala and the orbito-frontal cortex (OFC) in the processing of em...

متن کامل

Identification of emotional intonation evaluated by fMRI.

During acoustic communication among human beings, emotional information can be expressed both by the propositional content of verbal utterances and by the modulation of speech melody (affective prosody). It is well established that linguistic processing is bound predominantly to the left hemisphere of the brain. By contrast, the encoding of emotional intonation has been assumed to depend specif...

متن کامل

The neural response to emotional prosody, as revealed by functional magnetic resonance imaging.

Prosody is an important feature of language, comprising intonation, loudness, and tempo. Emotional prosodic processing forms an integral part of our social interactions. The main aim of this study was to use bold contrast fMRI to clarify the normal functional neuroanatomy of emotional prosody, in passive and active contexts. Subjects performed six separate scanning studies, within which two dif...

متن کامل

Cerebral pathways in processing of affective prosody: a dynamic causal modeling study.

This study was conducted to investigate the connectivity architecture of neural structures involved in processing of emotional speech melody (prosody). 24 subjects underwent event-related functional magnetic resonance imaging (fMRI) while rating the emotional valence of either prosody or semantics of binaurally presented adjectives. Conventional analysis of fMRI data revealed activation within ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2008